SERAPH: Semi-supervised Metric Learning Paradigm with Hyper Sparsity

نویسندگان

  • Gang Niu
  • Bo Dai
  • Makoto Yamada
  • Masashi Sugiyama
چکیده

We consider the problem of learning a distance metric from a limited amount of pairwise information as effectively as possible. The proposed SERAPH (SEmi-supervised metRic leArning Paradigm with Hyper sparsity) is a direct and substantially more natural approach for semi-supervised metric learning, since the supervised and unsupervised parts are based on a unified information theoretic framework. Unlike other extensions, the unsupervised part of SERAPH can extract further pairwise information from the unlabeled data according to temporary results of the supervised part, and therefore interacts with the supervised part positively. SERAPH involves both the sparsity of posterior distributions over the unobserved weak labels and the sparsity of the induced projection matrices, which we call the hyper sparsity. The resulting optimization is solved by an EM-like scheme, where the M-Step is convex, and the E-Step has analytical solution. Experimental results show that SERAPH compares favorably with existing metric learning algorithms based on weak labels.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Information-theoretic Semi-supervised Metric Learning via Entropy Regularization

We propose a general information-theoretic approach to semi-supervised metric learning called SERAPH (SEmi-supervised metRic leArning Paradigm with Hypersparsity) that does not rely on the manifold assumption. Given the probability parameterized by a Mahalanobis distance, we maximize its entropy on labeled data and minimize its entropy on unlabeled data following entropy regularization. For met...

متن کامل

Composite Kernel Optimization in Semi-Supervised Metric

Machine-learning solutions to classification, clustering and matching problems critically depend on the adopted metric, which in the past was selected heuristically. In the last decade, it has been demonstrated that an appropriate metric can be learnt from data, resulting in superior performance as compared with traditional metrics. This has recently stimulated a considerable interest in the to...

متن کامل

Some Research Problems in Metric Learning and Manifold Learning

In the past few years, metric learning, semi-supervised learning, and manifold learning methods have aroused a great deal of interest in the machine learning community. Many machine learning and pattern recognition algorithms rely on a distance metric. Instead of choosing the metric manually, a promising approach is to learn the metric from data automatically. Besides some early work on metric ...

متن کامل

Non-Negative Semi-Supervised Learning

The contributions of this paper are three-fold. First, we present a general formulation for reaping the benefits from both non-negative data factorization and semi-supervised learning, and the solution naturally possesses the characteristics of sparsity, robustness to partial occlusions, and greater discriminating power via extra unlabeled data. Then, an efficient multiplicative updating proced...

متن کامل

Adversarial Dropout for Supervised and Semi-supervised Learning

Recently, training with adversarial examples, which are generated by adding a small but worst-case perturbation on input examples, has improved the generalization performance of neural networks. In contrast to the biased individual inputs to enhance the generality, this paper introduces adversarial dropout, which is a minimal set of dropouts that maximize the divergence between 1) the training ...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:
  • CoRR

دوره abs/1105.0167  شماره 

صفحات  -

تاریخ انتشار 2011